CEQE: Contextualized Embeddings for Query Expansion

نویسندگان

چکیده

In this work we leverage recent advances in context-sensitive language models to improve the task of query expansion. Contextualized word representation models, such as ELMo and BERT, are rapidly replacing static embedding models. We propose a new model, Embeddings for Query Expansion (CEQE), that utilizes query-focused contextualized vectors. study behavior contextual representations generated expansion ad-hoc document retrieval. conduct our experiments on probabilistic retrieval well combination with neural ranking evaluate CEQE two standard TREC collections: Robust Deep Learning. find outperforms embedding-based methods multiple collections (by up 18% 31% Learning average precision) also improves over proven pseudo-relevance feedback (PRF) further passes reranking result continued gains effectiveness CEQE-based approaches outperforming other approaches. The final model incorporating score achieves 5% P@20 2% AP state-of-the-art transformer-based re-ranking Birch.

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Using Word Embeddings for Automatic Query Expansion

In this paper a framework for Automatic Query Expansion (AQE) is proposed using distributed neural language model word2vec. Using semantic and contextual relation in a distributed and unsupervised framework, word2vec learns a low dimensional embedding for each vocabulary entry. Using such a framework, we devise a query expansion technique, where related terms to a query are obtained by K-neares...

متن کامل

Query Expansion with Locally-Trained Word Embeddings

Continuous space word embeddings have received a great deal of attention in the natural language processing and machine learning communities for their ability to model term similarity and other relationships. We study the use of term relatedness in the context of query expansion for ad hoc information retrieval. We demonstrate that word embeddings such as word2vec and GloVe, when trained global...

متن کامل

Learning Concept Embeddings for Query Expansion by Quantum Entropy Minimization

In web search, users queries are formulated using only few terms and term-matching retrieval functions could fail at retrieving relevant documents. Given a user query, the technique of query expansion (QE) consists in selecting related terms that could enhance the likelihood of retrieving relevant documents. Selecting such expansion terms is challenging and requires a computational framework ca...

متن کامل

Mathematical Information Retrieval based on Type Embeddings and Query Expansion

We present an approach to mathematical information retrieval (MIR) that exploits a special kind of technical terminology, referred to as a mathematical type. In this paper, we present and evaluate a type detection mechanism and show its positive effect on the retrieval of research-level mathematics. Our best model, which performs query expansion with a type-aware embedding space, strongly outpe...

متن کامل

Combining fields for query expansion and adaptive query expansion

In this paper, we aim to improve query expansion for ad-hoc retrieval, by proposing a more fine-grained term reweighting process. This fine-grained process uses statistics from the representation of documents in various fields, such as their titles, the anchor text of their incoming links, and their body content. The contribution of this paper is twofold: First, we propose a novel query expansi...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Lecture Notes in Computer Science

سال: 2021

ISSN: ['1611-3349', '0302-9743']

DOI: https://doi.org/10.1007/978-3-030-72113-8_31